On Kernelizing Mahalanobis Distance Learning Algorithms
نویسندگان
چکیده
This paper focuses on the problem of kernelizing an existing supervised Mahalanobis distance learner. The following features are included in the paper. Firstly, three popular learners, namely, “neighborhood component analysis”, “large margin nearest neighbors” and “discriminant neighborhood embedding”, which do not have kernel versions are kernelized in order to improve their classification performances. Secondly, an alternative kernelization framework called “KPCA trick” is presented. Implementing a learner in the new framework gains several advantages over the standard framework. Thirdly, while the truths of representer theorems are just assumptions in previous papers related to ours, here, representer theorems are formally proven. The proofs validate both the kernel trick and the KPCA trick in the context of Mahalanobis distance learning. Fourthly, unlike previous works which always apply brute force methods to select a kernel, we investigate two approaches which can be efficiently adopted to construct an appropriate kernel for a given dataset. Finally, numerical results on various real-world datasets are presented to show the performances of the kernelized algorithms.
منابع مشابه
On Kernelization of Supervised Mahalanobis Distance Learners
This paper contains three contributions to the problem of learning a Mahalanobis distance. First, a general framework for kernelizing Mahalanobis distance learners is presented. The framework allows existing algorithms to learn a Mahalanobis distance in a feature space associated with a pre-specified kernel function. The framework is then used for kernelizing three well-known learners, namely, ...
متن کاملMetric and Kernel Learning Using a Linear Transformation
Metric and kernel learning arise in several machine learning applications. However, most existing metric learning algorithms are limited to learning metrics over low-dimensional data, while existing kernel learning algorithms are often limited to the transductive setting and do not generalize to new data points. In this paper, we study the connections between metric learning and kernel learning...
متن کاملFuzzy Possibility C-Mean Based on Mahalanobis Distance and Separable Criterion
The well known fuzzy partition clustering algorithms are most based on Euclidean distance function, which can only be used to detect spherical structural clusters. Gustafson-Kessel (GK) clustering algorithm and Gath-Geva (GG) clustering algorithm, were developed to detect non-spherical structural clusters, but both of them based on semi-supervised Mahalanobis distance needed additional prior in...
متن کاملA Metric-learning based framework for Support Vector Machines and Multiple Kernel Learning
Most metric learning algorithms, as well as Fisher’s Discriminant Analysis (FDA), optimize some cost function of different measures of within-and between-class distances. On the other hand, Support Vector Machines(SVMs) and several Multiple Kernel Learning (MKL) algorithms are based on the SVM large margin theory. Recently, SVMs have been analyzed from a metric learning perspective, and formula...
متن کاملLearning Distance Functions: Algorithms and Applications
This thesis presents research in the field of distance learning. Distance functions are extensively used in various application domains and also serve as an important building block in many types of algorithms. Despite their abundance, until recently only canonical distance functions such as the Euclidean distance have been used, or alternatively various application specific distance functions ...
متن کامل